latexit latexit sha1
Bellman Meets Hawkes: Model-Based Reinforcement Learning via Temporal Point Processes
Qu, Chao, Tan, Xiaoyu, Xue, Siqiao, Shi, Xiaoming, Zhang, James, Mei, Hongyuan
The last several years have witnessed the great success of reinforcement learning (RL) including the video game playing [Mnih et al., 2015], robot manipulation [Gu et al., 2017], autonomous driving [Shalev-Shwartz et al., 2016] and many others [Lazic et al., 2018, Dalal et al., 2016]. Most of them focus on the problem where the system of interest evolves continuously with time, e.g., a trajectory of a tennis ball. However, the conventional research in RL may omit a category of system that evolves continuously and may be interrupted by stochastic events abruptly (see the jumps in Figure 1). Such system exists ubiquitously in the social and information science and therefore necessitates the research of reinforcement learning in these domains to extend its applicability in the real-world problems [Farajtabar et al., 2017, Wang et al., 2018], in which the agent seeks an optimal intervention policy so as to improve the future course of events. Concrete examples may include: - Social media. Social media website allows users to create and share content. Retweet can form as users resharing and broadcasting others' tweet to their friends and followers. Such stochastic events would steer the behaviors of other tweet users [Rizoiu et al., 2017]. At the same time, the platform (agent) may want to seek a policy to effectively mitigate the fake news by optimizing the performance of real news propagation over the network Farajtabar et al. [2017].
- North America > United States > Illinois > Cook County > Chicago (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- Research Report (0.82)
- Instructional Material > Course Syllabus & Notes (0.46)
- Energy (0.92)
- Leisure & Entertainment > Sports > Tennis (0.74)
- Information Technology > Services (0.67)
- Leisure & Entertainment > Games > Computer Games (0.48)
Directed Weight Neural Networks for Protein Structure Representation Learning
Li, Jiahan, Luo, Shitong, Deng, Congyue, Cheng, Chaoran, Guan, Jiaqi, Guibas, Leonidas, Peng, Jian, Ma, Jianzhu
A protein performs biological functions by folding to a particular 3D structure. To accurately model the protein structures, both the overall geometric topology and local fine-grained relations between amino acids (e.g. side-chain torsion angles and inter-amino-acid orientations) should be carefully considered. In this work, we propose the Directed Weight Neural Network for better capturing geometric relations among different amino acids. Extending a single weight from a scalar to a 3D directed vector, our new framework supports a rich set of geometric operations on both classical and SO(3)--representation features, on top of which we construct a perceptron unit for processing amino-acid information. In addition, we introduce an equivariant message passing paradigm on proteins for plugging the directed weight perceptrons into existing Graph Neural Networks, showing superior versatility in maintaining SO(3)-equivariance at the global scale. Experiments show that our network has remarkably better expressiveness in representing geometric relations in comparison to classical neural networks and the (globally) equivariant networks. It also achieves state-of-the-art performance on various computational biology applications related to protein 3D structures.
- Europe > Austria > Vienna (0.14)
- North America > United States > Illinois (0.04)
- North America > United States > Colorado (0.04)
- Asia > China (0.04)
One-Shot Learning on Attributed Sequences
Zhuang, Zhongfang, Kong, Xiangnan, Rundensteiner, Elke, Arora, Aditya, Zouaoui, Jihane
One-shot learning has become an important research topic in the last decade with many real-world applications. The goal of one-shot learning is to classify unlabeled instances when there is only one labeled example per class. Conventional problem setting of one-shot learning mainly focuses on the data that is already in feature space (such as images). However, the data instances in real-world applications are often more complex and feature vectors may not be available. In this paper, we study the problem of one-shot learning on attributed sequences, where each instance is composed of a set of attributes (e.g., user profile) and a sequence of categorical items (e.g., clickstream). This problem is important for a variety of real-world applications ranging from fraud prevention to network intrusion detection. This problem is more challenging than conventional one-shot learning since there are dependencies between attributes and sequences. We design a deep learning framework OLAS to tackle this problem. The proposed OLAS utilizes a twin network to generalize the features from pairwise attributed sequence examples. Empirical results on real-world datasets demonstrate the proposed OLAS can outperform the state-of-the-art methods under a rich variety of parameter settings.
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- Asia > Middle East > Jordan (0.04)
- Information Technology > Security & Privacy (0.48)
- Health & Medicine (0.46)
Deep Learning on Attributed Sequences
Recent research in feature learning has been extended to sequence data, where each instance consists of a sequence of heterogeneous items with a variable length. However, in many real-world applications, the data exists in the form of attributed sequences, which is composed of a set of fixed-size attributes and variable-length sequences with dependencies between them. In the attributed sequence context, feature learning remains challenging due to the dependencies between sequences and their associated attributes. In this dissertation, we focus on analyzing and building deep learning models for four new problems on attributed sequences. Our extensive experiments on real-world datasets demonstrate that the proposed solutions significantly improve the performance of each task over the state-of-the-art methods on attributed sequences.
- Information Technology > Security & Privacy (0.92)
- Law Enforcement & Public Safety (0.67)
- Education > Educational Setting > Online (0.45)
ICLEA: Interactive Contrastive Learning for Self-supervised Entity Alignment
Zeng, Kaisheng, Dong, Zhenhao, Hou, Lei, Cao, Yixin, Hu, Minghao, Yu, Jifan, Lv, Xin, Li, Juanzi, Feng, Ling
Self-supervised entity alignment (EA) aims to link equivalent entities across different knowledge graphs (KGs) without seed alignments. The current SOTA self-supervised EA method draws inspiration from contrastive learning, originally designed in computer vision based on instance discrimination and contrastive loss, and suffers from two shortcomings. Firstly, it puts unidirectional emphasis on pushing sampled negative entities far away rather than pulling positively aligned pairs close, as is done in the well-established supervised EA. Secondly, KGs contain rich side information (e.g., entity description), and how to effectively leverage those information has not been adequately investigated in self-supervised EA. In this paper, we propose an interactive contrastive learning model for self-supervised EA. The model encodes not only structures and semantics of entities (including entity name, entity description, and entity neighborhood), but also conducts cross-KG contrastive learning by building pseudo-aligned entity pairs. Experimental results show that our approach outperforms previous best self-supervised results by a large margin (over 9% average improvement) and performs on par with previous SOTA supervised counterparts, demonstrating the effectiveness of the interactive contrastive learning for self-supervised EA.
Glance and Focus Networks for Dynamic Visual Recognition
Huang, Gao, Wang, Yulin, Lv, Kangchen, Jiang, Haojun, Huang, Wenhui, Qi, Pengfei, Song, Shiji
Spatial redundancy widely exists in visual recognition tasks, i.e., discriminative features in an image or video frame usually correspond to only a subset of pixels, while the remaining regions are irrelevant to the task at hand. Therefore, static models which process all the pixels with an equal amount of computation result in considerable redundancy in terms of time and space consumption. In this paper, we formulate the image recognition problem as a sequential coarse-to-fine feature learning process, mimicking the human visual system. Specifically, the proposed Glance and Focus Network (GFNet) first extracts a quick global representation of the input image at a low resolution scale, and then strategically attends to a series of salient (small) regions to learn finer features. The sequential process naturally facilitates adaptive inference at test time, as it can be terminated once the model is sufficiently confident about its prediction, avoiding further redundant computation. It is worth noting that the problem of locating discriminant regions in our model is formulated as a reinforcement learning task, thus requiring no additional manual annotations other than classification labels. GFNet is general and flexible as it is compatible with any off-the-shelf backbone models (such as MobileNets, EfficientNets and TSM), which can be conveniently deployed as the feature extractor. Extensive experiments on a variety of image classification and video recognition tasks and with various backbone models demonstrate the remarkable efficiency of our method. For example, it reduces the average latency of the highly efficient MobileNet-V3 on an iPhone XS Max by 1.3x without sacrificing accuracy. Code and pre-trained models are available at https://github.com/blackfeather-wang/GFNet-Pytorch.
- Asia > China > Beijing > Beijing (0.05)
- Asia > China > Heilongjiang Province > Harbin (0.04)
- North America > United States > New York > Tompkins County > Ithaca (0.04)
- North America > United States > California > Alameda County > Berkeley (0.04)
- Information Technology (0.46)
- Telecommunications (0.46)
Smooth Nested Simulation: Bridging Cubic and Square Root Convergence Rates in High Dimensions
Wang, Wenjia, Wang, Yanyuan, Zhang, Xiaowei
Nested simulation concerns estimating functionals of a conditional expectation via simulation. In this paper, we propose a new method based on kernel ridge regression to exploit the smoothness of the conditional expectation as a function of the multidimensional conditioning variable. Asymptotic analysis shows that the proposed method can effectively alleviate the curse of dimensionality on the convergence rate as the simulation budget increases, provided that the conditional expectation is sufficiently smooth. The smoothness bridges the gap between the cubic root convergence rate (that is, the optimal rate for the standard nested simulation) and the square root convergence rate (that is, the canonical rate for the standard Monte Carlo simulation). We demonstrate the performance of the proposed method via numerical examples from portfolio risk management and input uncertainty quantification.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Asia > Japan > Honshū > Kantō > Kanagawa Prefecture (0.04)
- Asia > China > Hong Kong (0.04)
- (2 more...)
AdaFocus V2: End-to-End Training of Spatial Dynamic Networks for Video Recognition
Wang, Yulin, Yue, Yang, Lin, Yuanze, Jiang, Haojun, Lai, Zihang, Kulikov, Victor, Orlov, Nikita, Shi, Humphrey, Huang, Gao
Recent works have shown that the computational efficiency of video recognition can be significantly improved by reducing the spatial redundancy. As a representative work, the adaptive focus method (AdaFocus) has achieved a favorable trade-off between accuracy and inference speed by dynamically identifying and attending to the informative regions in each video frame. However, AdaFocus requires a complicated three-stage training pipeline (involving reinforcement learning), leading to slow convergence and is unfriendly to practitioners. This work reformulates the training of AdaFocus as a simple one-stage algorithm by introducing a differentiable interpolation-based patch selection operation, enabling efficient end-to-end optimization. We further present an improved training scheme to address the issues introduced by the one-stage formulation, including the lack of supervision, input diversity and training stability. Moreover, a conditional-exit technique is proposed to perform temporal adaptive computation on top of AdaFocus without additional training. Extensive experiments on six benchmark datasets (i.e., ActivityNet, FCVID, Mini-Kinetics, Something-Something V1&V2, and Jester) demonstrate that our model significantly outperforms the original AdaFocus and other competitive baselines, while being considerably more simple and efficient to train. Code is available at https://github.com/LeapLabTHU/AdaFocusV2.
- North America > United States > Oregon (0.04)
- Asia > Middle East > Qatar > Ad-Dawhah > Doha (0.04)
- Asia > China > Beijing > Beijing (0.04)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (0.66)
CausalMTA: Eliminating the User Confounding Bias for Causal Multi-touch Attribution
Yao, Di, Gong, Chang, Zhang, Lei, Chen, Sheng, Bi, Jingping
Multi-touch attribution (MTA), aiming to estimate the contribution of each advertisement touchpoint in conversion journeys, is essential for budget allocation and automatically advertising. Existing methods first train a model to predict the conversion probability of the advertisement journeys with historical data and calculate the attribution of each touchpoint using counterfactual predictions. An assumption of these works is the conversion prediction model is unbiased, i.e., it can give accurate predictions on any randomly assigned journey, including both the factual and counterfactual ones. Nevertheless, this assumption does not always hold as the exposed advertisements are recommended according to user preferences. This confounding bias of users would lead to an out-of-distribution (OOD) problem in the counterfactual prediction and cause concept drift in attribution. In this paper, we define the causal MTA task and propose CausalMTA to eliminate the influence of user preferences. It systemically eliminates the confounding bias from both static and dynamic preferences to learn the conversion prediction model using historical data. We also provide a theoretical analysis to prove CausalMTA can learn an unbiased prediction model with sufficient data. Extensive experiments on both public datasets and the impression data in an e-commerce company show that CausalMTA not only achieves better prediction performance than the state-of-the-art method but also generates meaningful attribution credits across different advertising channels.
- North America > United States > California > San Diego County > San Diego (0.04)
- North America > Canada > Quebec > Montreal (0.04)
- North America > Canada > Nova Scotia > Halifax Regional Municipality > Halifax (0.04)
- (5 more...)
- Research Report > Promising Solution (0.48)
- Research Report > Experimental Study (0.34)
- Marketing (1.00)
- Information Technology > Services (0.67)
A random energy approach to deep learning
Xie, Rongrong, Marsili, Matteo
The study of ensembles of random systems can provide several insights on the properties of complex systems, such as heavy ions [29], ecologies [15], disordered materials [17], satisfiability in computer science [18] and machine learning [30]. Indeed the collective behaviour of a system composed of many interacting degrees of freedom often does not depend on the specific realisation of the wiring of the interactions, but only on the statistical properties of the resulting energy landscape. In these circumstances, any realisation of a random system that shares the same statistical properties enjoys the same "typical" collective behaviour. The Random Energy Model (REM) [6] is probably the simplest exemplar of this approach. It makes minimal assumptions on the network of interactions, because interactions of any order can occur among the variables [6]. It features a phase transition between a random (high temperature) phase and a low temperature frozen phase, which reproduces the gross features of more complex systems such as spin glasses.
- North America > United States > New York (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- Europe > Italy > Friuli Venezia Giulia > Trieste Province > Trieste (0.04)
- Asia > China > Hubei Province > Wuhan (0.04)